New data-parallel language features for sparse matrix computations
نویسندگان
چکیده
High-level data-parallel languages such as Vienna Fortran and High Performance Fortran (HPF) have been introduced to allow the programming of massively parallel distributed-memory machines at a relatively high level of abstraction, based on the Single-Program-Multiple-Data (SPMD) paradigm. Their main features include mechanisms for expressing the distribution of data across the processors of a machine. This paper introduces additional language functionality to allow the eecient processing of sparse matrix codes. We introduce new methods for the representation and distribution of sparse matrices, which forms a powerful mechanism for storing and manipulating sparse matrices able to be eeciently implemented on massively parallel machines.
منابع مشابه
HPF-2 Support for Dynamic Sparse Computations
There is a class of sparse matrix computations, such as direct solvers of systems of linear equations, that change the fill-in (nonzero entries) of the coefficient matrix, and involve row and column operations (pivoting). This paper addresses the problem of the parallelization of these sparse computations from the point of view of the parallel language and the compiler. Dynamic data structures ...
متن کاملVienna-Fortran/HPF Extensions for Sparse and Irregular Problems and Their Compilation
Vienna Fortran, High Performance Fortran (HPF), and other data parallel languages have been introduced to allow the programming of massively parallel distributed-memory machines (DMMP) at a relatively high level of abstraction, based on the SPMD paradigm. Their main features include directives to express the distribution of data and computations across the processors of a machine. In this paper...
متن کاملSupport for Irregular Computations in Massively Parallel PIM Arrays, Using an Object-Based Execution Model
The emergence of semiconductor fabrication technology allowing a tight coupling between high-density DRAM and CMOS logic on the same chip has led to the important new class of Processor-In-Memory (PIM) architectures. Furthermore, large arrays of PIMs can be arranged into massively parallel architectures. In this paper, we outline the salient features of PIM architectures and discuss macroserver...
متن کاملDesigning Parallel Sparse Matrix Algorithms beyond Data Dependence Analysis
Algorithms are often parallelized based on data dependence analysis manually or by means of parallel compilers. Some vector/matrix computations such as the matrix-vector products with simple data dependence structures (data parallelism) can be easily parallelized. For problems with more complicated data dependence structures, parallelization is less straightforward. The data dependence graph is...
متن کاملTime Complexity and Parallel Speedup to Compute the Gamma Summarization Matrix
We study the serial and parallel computation of Γ (Gamma), a comprehensive data summarization matrix for linear Gaussian models, widely used in big data analytics. Computing Gamma can be reduced to a single matrix multiplication with the data set, where such multiplication can be evaluated as a sum of vector outer products, which enables incremental and parallel computation, essential features ...
متن کامل